Contrastive Divergence in Gaussian Diffusions

نویسنده

  • Javier R. Movellan
چکیده

This letter presents an analysis of the contrastive divergence (CD) learning algorithm when applied to continuous-time linear stochastic neural networks. For this case, powerful techniques exist that allow a detailed analysis of the behavior of CD. The analysis shows that CD converges to maximum likelihood solutions only when the network structure is such that it can match the first moments of the desired distribution. Otherwise, CD can converge to solutions arbitrarily different from the log-likelihood solutions, or they can even diverge. This result suggests the need to improve our theoretical understanding of the conditions under which CD is expected to be well behaved and the conditions under which it may fail. In, addition the results point to practical ideas on how to improve the performance of CD.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Gaussian Approximations of Small Noise Diffusions in Kullback-leibler Divergence

Abstract. We study Gaussian approximations to the distribution of a diffusion. The approximations are easy to compute: they are defined by two simple ordinary differential equations for the mean and the covariance. Time correlations can also be computed via solution of a linear stochastic differential equation. We show, using the Kullback-Leibler divergence, that the approximations are accurate...

متن کامل

Fast Communication Gaussian Approximations of Small Noise Diffusions in Kullback–leibler Divergence∗

We study Gaussian approximations to the distribution of a diffusion. The approximations are easy to compute: they are defined by two simple ordinary differential equations for the mean and the covariance. Time correlations can also be computed via solution of a linear stochastic differential equation. We show, using the Kullback–Leibler divergence, that the approximations are accurate in the sm...

متن کامل

An Analysis of Contrastive Divergence Learning in Gaussian Boltzmann Machines

The Boltzmann machine (BM) learning rule for random field models with latent variables can be problematic to use in practice. These problems have (at least partially) been attributed to the negative phase in BM learning where a Gibbs sampling chain should be run to equilibrium. Hinton (1999, 2000) has introduced an alternative called contrastive divergence (CD) learning where the chain is run f...

متن کامل

Contrastive Divergence In Gaussian Diffusion Processes

Contrastive Divergence (CD) is a recent learning rule found to work very well in practice despite its still unclear theoretical underpinnings (Carreira-Perpiñan and Hinton, 2005; Hinton, 2002; Hinton and Salakhutdinov, 2006; Hyvärinen, 2006; MacKay, 2001; Roth and Black, 2005; Williams and Agakov, 2002; Yuille, 2004). This paper presents a theoretical analysis of CD in Gaussian Diffusion Proces...

متن کامل

Differential Contrastive Divergence

We formulate a differential version of contrastive divergence for continuous configuration spaces by considering a limit of MCMC processes in which the proposal distribution becomes infinitesimal. This leads to a deterministic differential contrastive divergence update — one in which no stochastic sampling is required. We prove convergence of differential contrastive divergence in general and p...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Neural computation

دوره 20 9  شماره 

صفحات  -

تاریخ انتشار 2008